Skip to content

[ROCm][CI] Add large_gpu_mark to test_max_tokens_none for ROCm#37717

Merged
DarkLight1337 merged 1 commit intovllm-project:mainfrom
ROCm:akaratza_fix_regression_test
Mar 22, 2026
Merged

[ROCm][CI] Add large_gpu_mark to test_max_tokens_none for ROCm#37717
DarkLight1337 merged 1 commit intovllm-project:mainfrom
ROCm:akaratza_fix_regression_test

Conversation

@AndreasKaratzas
Copy link
Copy Markdown
Collaborator

Follow-up for:

Marks max_tokens test with distilbert/distilgpt2 as a large GPU test. Addresses failure in mi250_1: Regression

Motivation: https://buildkite.com/vllm/amd-ci/builds/6721/steps/canvas?sid=019d09d4-708e-44b0-a0d0-ccf0e3c00a94&tab=output

cc @kenroche

Signed-off-by: Andreas Karatzas <akaratza@amd.com>
@AndreasKaratzas AndreasKaratzas added ready ONLY add when PR is ready to merge/full CI is needed rocm Related to AMD ROCm labels Mar 20, 2026
@github-project-automation github-project-automation bot moved this to Todo in AMD Mar 20, 2026
@AndreasKaratzas
Copy link
Copy Markdown
Collaborator Author

Testing MI250 to see if issue is resolved (added rocm and ready labels).

Copy link
Copy Markdown
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

The pull request effectively addresses the CI failure on ROCm platforms by introducing the large_gpu_mark to the test_max_tokens_none test. This ensures that the test is only run on environments with sufficient GPU memory, improving the reliability and accuracy of the test suite. The implementation is clean and correctly leverages pytest.mark.parametrize for conditional marking based on the platform.

@AndreasKaratzas
Copy link
Copy Markdown
Collaborator Author

@AndreasKaratzas AndreasKaratzas marked this pull request as ready for review March 21, 2026 00:57
Comment thread tests/test_regression.py
"model",
[
pytest.param(
"distilbert/distilgpt2",
Copy link
Copy Markdown
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This model only has 88.2M params, how does it OOM?

Copy link
Copy Markdown
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can you add it to #37736?

@DarkLight1337 DarkLight1337 merged commit c86b17c into vllm-project:main Mar 22, 2026
16 checks passed
@github-project-automation github-project-automation bot moved this from Todo to Done in AMD Mar 22, 2026
@AndreasKaratzas AndreasKaratzas deleted the akaratza_fix_regression_test branch March 22, 2026 04:29
RhizoNymph pushed a commit to RhizoNymph/vllm that referenced this pull request Mar 26, 2026
SouthWest7 pushed a commit to SouthWest7/vllm that referenced this pull request Mar 27, 2026
khairulkabir1661 pushed a commit to khairulkabir1661/vllm that referenced this pull request Mar 27, 2026
nithinvc pushed a commit to nithinvc/vllm that referenced this pull request Mar 27, 2026
…project#37717)

Signed-off-by: Andreas Karatzas <akaratza@amd.com>

Signed-off-by: Nithin Chalapathi <nithin.ch10@gmail.com>
JiantaoXu pushed a commit to JiantaoXu/vllm that referenced this pull request Mar 28, 2026
mtparet pushed a commit to blackfuel-ai/vllm that referenced this pull request Apr 9, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ready ONLY add when PR is ready to merge/full CI is needed rocm Related to AMD ROCm

Projects

Status: Done

Development

Successfully merging this pull request may close these issues.

2 participants